# Whole Word Masking Pre-training
Nezha Base Wwm
NEZHA is a Chinese pre-trained language model based on the Transformer architecture, optimized for Chinese text understanding tasks using the whole word masking strategy
Large Language Model
Transformers

N
sijunhe
66
2
Chinese Bert Wwm Ext
Apache-2.0
A Chinese pre-trained BERT model employing whole word masking strategy, aimed at accelerating Chinese natural language processing research.
Large Language Model Chinese
C
hfl
24.49k
174
Rbtl3
Apache-2.0
This is a retrained three-layer RoBERTa-wwm-ext-large model, a Chinese pre-trained BERT model employing whole word masking strategy, aimed at accelerating the development of Chinese natural language processing.
Large Language Model Chinese
R
hfl
767
4
Rbt3
Apache-2.0
This is a Chinese pre-trained BERT model employing whole word masking technology, developed by the HIT-iFLYTEK Joint Lab to accelerate advancements in Chinese natural language processing.
Large Language Model Chinese
R
hfl
6,626
35
Bert Base Arabic Camelbert Da
Apache-2.0
CAMeLBERT-DA is a BERT model pre-trained on Dialectal Arabic (DA), part of the CAMeLBERT series, specializing in Arabic NLP tasks.
Large Language Model Arabic
B
CAMeL-Lab
1,756
28
Featured Recommended AI Models